# Russian optimization
MT Gemma 3 12B
This project uses the mergekit and DARE TIES methods to merge the soob3123/amoral-gemma3-12B-v2 and IlyaGusev/saiga_gemma3_12b models, aiming to provide more powerful language processing capabilities.
Large Language Model
Transformers

M
zelk12
1,348
2
Ruadaptqwen2.5 7B Lite Beta
Apache-2.0
This project adapts the T-lite-it-1.0 model to Russian. By replacing the tokenizer, continuing pre-training on the Russian corpus, and applying LEP technology, the generation speed of Russian text is improved.
Large Language Model
Safetensors
R
RefalMachine
1,603
8
Ruadaptqwen2.5 32B Pro Beta
Apache-2.0
This project is the result of adapting the model to Russian. By replacing the tokenizer, continuing pre-training on the Russian corpus, and applying LEP technology, the generation speed of Russian text has been improved, providing a new solution for Russian natural language processing.
Large Language Model
Safetensors
R
RefalMachine
2,913
10
Meno Tiny 0.1
Apache-2.0
A Russian instruction-optimized model fine-tuned based on Qwen2.5-1.5B-Instruct, with 1.5B parameters, supporting both Russian and English, excelling in text QA within RAG workflows
Large Language Model
Transformers Supports Multiple Languages

M
bond005
764
10
Vikhr Llama 3.2 1B Instruct
A Russian-optimized instruction model based on Llama-3.2-1B-Instruct, with a 5-fold increase in efficiency, suitable for deployment on low-power devices
Large Language Model
Transformers

V
Vikhrmodels
2,915
41
It 5.3 Fp16 32k
Apache-2.0
Cyclone 0.5* is an open-source large language model supporting Russian and English, extending the RoPE context window to 32k and optimizing JSON processing and multi-turn dialogue capabilities
Large Language Model
Transformers Supports Multiple Languages

I
Vikhrmodels
74
11
Vikhr 7b 0.1
Apache-2.0
Vikhr is a Russian language model based on the Mistral architecture, undergoing three rounds of pre-training with 400 million tokens. It outperforms the original Mistral in Russian tasks but may have shortcomings in code processing.
Large Language Model
Transformers Supports Multiple Languages

V
Vikhrmodels
259
57
Rut5 Base
MIT
A streamlined version based on google/mt5-base, optimized for Russian and English with 58% fewer parameters
Large Language Model Supports Multiple Languages
R
cointegrated
27.85k
11
Featured Recommended AI Models